# Information Retrieval Optimization

Finetuned Cross Encoder L6 V2
This is a fine-tuned cross-encoder model based on cross-encoder/ms-marco-MiniLM-L6-v2, primarily used for text re-ranking and semantic search tasks.
Text Embedding
F
CharlesPing
22
1
GTE ModernColBERT V1
Apache-2.0
PyLate is a sentence similarity model based on the ColBERT architecture, using Alibaba-NLP/gte-modernbert-base as the base model and trained with distillation loss, suitable for information retrieval tasks.
Text Embedding
G
lightonai
157.96k
98
Reranker ModernBERT Base Gooaq Bce
Apache-2.0
This is a cross-encoder model fine-tuned from ModernBERT-base for text re-ranking and semantic search tasks.
Text Embedding English
R
akr2002
16
1
Reasoning Bert Ccnews
This is a fine-tuned BERT-based sentence transformer model for mapping sentences and paragraphs into a 768-dimensional vector space, supporting tasks such as semantic text similarity and semantic search.
Text Embedding
R
bwang0911
13
1
Reranker Bert Tiny Gooaq Bce Tanh V4
Apache-2.0
This is a cross-encoder model fine-tuned from bert-tiny for computing similarity scores between text pairs, suitable for tasks like semantic textual similarity and semantic search.
Text Embedding English
R
cross-encoder-testing
1,971
0
Rank1 32b
MIT
rank1-32b is an information retrieval reranking model based on Qwen2.5-32B, which judges relevance by generating reasoning chains
Large Language Model Transformers English
R
jhu-clsp
18
0
Rank1 14b
MIT
rank1 is a 14-billion-parameter reasoning re-ranking model that improves the performance of information retrieval tasks by generating explicit reasoning chains before making relevance judgments.
Large Language Model Transformers English
R
jhu-clsp
23
0
Namaa ARA Reranker V1
Apache-2.0
A model specifically designed for Arabic reranking tasks, capable of accurately evaluating the relevance between queries and passages.
Text Embedding Transformers Arabic
N
NAMAA-Space
56
4
Arabic Reranker V1
This is an Arabic re-ranking model based on the BERT architecture, optimized for Arabic text relevance ranking tasks
Text Embedding Arabic
A
oddadmix
21
1
Llm2vec Meta Llama 31 8B Instruct Mntp
MIT
LLM2Vec is a simple method to convert decoder-only large language models into text encoders by enabling bidirectional attention, masked next-token prediction, and unsupervised contrastive learning.
Text Embedding Transformers English
L
McGill-NLP
386
2
Thusinh1969 Gemma2 2b Rerank Checkpoint 8800 Gguf
Text ranking model based on Gemma 2B architecture, offering multiple quantization versions to suit different hardware needs
T
RichardErkhov
71
0
Ruri Reranker Large
Apache-2.0
Ruri Reranker is a general-purpose Japanese reranking model based on the Sentence Transformers architecture, specifically designed for Japanese text relevance ranking tasks.
Text Embedding Japanese
R
cl-nagoya
2,538
11
Ruri Reranker Base
Apache-2.0
General-purpose Japanese reranking model for improving relevance ranking in Japanese text retrieval
Text Embedding Japanese
R
cl-nagoya
1,100
4
Ruri Reranker Stage1 Base
Apache-2.0
Ruri Reranker is a Japanese text reranking model based on Transformer architecture, specifically designed to optimize the ranking quality of retrieval results.
Text Embedding Japanese
R
cl-nagoya
26
0
Ruri Reranker Small
Apache-2.0
Ruri-Reranker is a reranking model specifically optimized for Japanese text, based on the sentence-transformers architecture, effectively improving the relevance ranking of search results.
Text Embedding Japanese
R
cl-nagoya
116
2
Ruri Reranker Stage1 Small
Apache-2.0
The Ruri Reranker is a general-purpose Japanese reranking model specifically designed to improve the relevance ranking of Japanese text retrieval results. The small version maintains high performance while having a smaller parameter count.
Text Embedding Japanese
R
cl-nagoya
25
0
Ko Reranker 8k
Apache-2.0
A text ranking model fine-tuned with Korean data based on BAAI/bge-reranker-v2-m3
Text Embedding Transformers Supports Multiple Languages
K
upskyy
14
11
Fingumv3
This is a sentence-transformers model fine-tuned from dunzhang/stella_en_1.5B_v5, designed to generate 1024-dimensional dense vector representations for sentences and paragraphs, suitable for tasks like semantic text similarity and semantic search.
Text Embedding
F
FINGU-AI
26
1
Monoelectra Base
Apache-2.0
lightning-ir is a cross-encoder model based on the ELECTRA architecture, specifically designed for text ranking tasks. The model optimizes paragraph reordering performance through large language model distillation techniques.
Large Language Model
M
webis
69
4
Crossencoder Xlm Roberta Base Mmarcofr
MIT
This is a French cross-encoder model based on XLM-RoBERTa, specifically designed for passage re-ranking tasks in semantic search.
Text Embedding French
C
antoinelouis
51
0
Venusaur
MIT
Venusaur is a sentence embedding model developed based on the Mihaiii/Bulbasaur foundation model, focusing on sentence similarity and feature extraction tasks.
Text Embedding
V
Mihaiii
290
3
Japanese Reranker Cross Encoder Small V1
MIT
This is a Japanese-trained Reranker (Cross-Encoder) model for text ranking tasks.
Text Embedding Japanese
J
hotchpotch
209
3
Instructor Xl
Apache-2.0
A T5 architecture-based sentence embedding model focused on semantic similarity and information retrieval tasks for English text.
Text Embedding Transformers English
I
retrainai
22
0
Simcse Small E Czech
A Czech sentence similarity model fine-tuned with SimCSE objective based on Seznam/small-e-czech model
Text Embedding Transformers Other
S
Seznam
1,543
1
Gte Small
MIT
GTE-small is a general text embedding model trained by Alibaba DAMO Academy, based on the BERT framework, suitable for tasks such as information retrieval and semantic text similarity.
Text Embedding Transformers English
G
Supabase
481.27k
89
Gte Small
MIT
GTE-small is a compact general-purpose text embedding model suitable for various natural language processing tasks, including sentence similarity calculation, text classification, and retrieval.
Text Embedding English
G
thenlper
450.86k
158
Gte Base
MIT
GTE-Base is a general-purpose text embedding model focused on sentence similarity and text retrieval tasks, performing well on multiple benchmarks.
Text Embedding English
G
thenlper
317.05k
117
Instructor Large Safetensors
Apache-2.0
INSTRUCTOR is a text embedding model based on the T5 architecture, focusing on sentence similarity calculation and information retrieval tasks. It excels in various NLP tasks, including text classification, clustering, and semantic similarity evaluation.
Text Embedding Transformers English
I
gentlebowl
16
0
Bart Ranker
MIT
This model is used to predict the relevance of query-document pairs, suitable for information retrieval tasks.
Text Embedding Transformers
B
bsl
31
3
Doc2query T5 Base Msmarco
A document expansion model based on T5-base architecture, trained on the MS MARCO dataset, used to generate potential queries related to document content to enhance retrieval effectiveness
Text Embedding Transformers English
D
macavaney
341
2
Mdpr Tied Pft Msmarco Ft All
This model is a dense retrieval model further fine-tuned on all Mr. TyDi training data based on the castorini/mdpr-tied-pft-msmarco checkpoint.
Large Language Model Transformers
M
castorini
386
0
Dense Encoder Distilbert Frozen Emb
Dense retrieval model based on DistilBERT architecture, trained on the MS MARCO dataset with frozen word embedding layers
Text Embedding Transformers
D
vocab-transformers
26
0
Doc2query T5 Base Msmarco
A retrieval model that converts documents into queries to enhance document search relevance
Large Language Model
D
castorini
1,064
14
Duot5 Base Msmarco
A text re-ranking model based on the T5-base architecture, fine-tuned on the MS MARCO passage dataset to improve the relevance ranking of information retrieval results.
Large Language Model
D
castorini
4,915
0
Msmarco Distilbert Word2vec256k MLM 230k
This model is a pre-trained language model based on the DistilBERT architecture, initialized with a 256k vocabulary using word2vec and trained on the MS MARCO corpus with masked language modeling (MLM).
Large Language Model Transformers
M
vocab-transformers
16
0
Monot5 3b Msmarco
A re-ranker based on the T5-3B architecture, fine-tuned for 100,000 steps on the MS MARCO passage dataset for document ranking tasks.
Large Language Model Transformers
M
castorini
737
0
Monot5 Base Msmarco
A re-ranking model based on the T5-base architecture, fine-tuned for 100,000 steps on the MS MARCO passage dataset, suitable for document re-ranking tasks in information retrieval.
Large Language Model
M
castorini
7,405
11
S PubMedBert MS MARCO
A sentence-transformers model fine-tuned on the MS-MARCO dataset based on PubMedBERT, suitable for semantic similarity calculation and information retrieval tasks in the medical/health text domain
Text Embedding Transformers
S
pritamdeka
30.50k
28
Bert Fa Base Uncased Wikinli Mean Tokens
Apache-2.0
A Persian sentence embedding model based on ParsBERT for generating high-quality sentence vector representations
Text Embedding Other
B
m3hrdadfi
555
0
Bert Base Msmarco
A fine-tuned version based on the BERT-Base architecture for the MS MARCO passage classification task, suitable for document re-ranking tasks
Large Language Model
B
Capreolus
64
0
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase